AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Efficient CPU Inference

# Efficient CPU Inference

Bge Large Zh V1.5 Gguf
MIT
BAAI/bge-large-zh-v1.5 embedding models in GGUF format, both quantized and non-quantized, optimized for llama.cpp, delivering significant speedup on CPU with minimal precision loss.
Text Embedding
B
CompendiumLabs
1,213
12
Bge Base En V1.5 Gguf
MIT
This project provides the BGE embedding model stored in GGUF format, which is suitable for use with llama.cpp and offers better performance than transformers.
Text Embedding
B
CompendiumLabs
1,108
5
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase